Guarding against Adversarial Domain Shifts

ثبت نشده
چکیده

When training a deep neural network for supervised image classification, one can broadly distinguish between two types of latent features of images that will drive the classification: (i) ‘immutable’ or ‘core’ features that are really inherent to the object in question and do not change substantially from one instance of the object to another and (ii) ‘mutable’ or ‘style’ features such as position, rotation, image quality or brightness but also more complex ones like hair color or posture for images of persons. The distribution of the style features can change in the future. While transfer learning and domain adaptation would try to adapt to a shift in the distribution(s), we here want to protect against future adversarial domain shifts, arising through changing style features, by ideally not using the mutable style features altogether. There are two broad scenarios and we show how exploiting grouping information in the data helps in both. (a) If the style features are known explicitly (such as translation, rotation, etc.) one usually proceeds by using data augmentation. By exploiting the grouping information about which original image an augmented sample belongs to, we can reduce the sample size required to achieve invariance to the style feature in question. (b) Sometimes the style features are not known explicitly but we still have information about samples that belong to the same underlying object (such as pictures of the same person in different circumstances). By constraining the classification to give the same forecast for all instances that belong to the same object, we show how using this grouping information leads to invariance to such implicit style features and helps to protect against adversarial domain shifts. We provide a causal framework for the problem and treat groups of instances of the same object as counterfactuals under different interventions on the mutable style features. We show links to questions of interpretability, fairness, transfer learning and adversarial examples.

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Grouping-By-ID: Guarding Against Adversarial Domain Shifts

When training a deep neural network for supervised image classification, one can broadly distinguish between two types of latent features of images that will drive the classification: (i) ‘immutable’ or ‘core’ features that are really inherent to the object in question and do not change substantially from one instance of the object to another and (ii) ‘mutable’ or ‘style’ features such as posit...

متن کامل

Incremental Adversarial Domain Adaptation

Continuous appearance shifts such as changes in weather and lighting conditions can impact the performance of deployed machine learning models. Unsupervised domain adaptation aims to address this challenge, though current approaches do not utilise the continuity of the occurring shifts. Many robotic applications exhibit these conditions and thus facilitate the potential to incrementally adapt a...

متن کامل

Grouping-by-id: Guarding against Adversar-

When training a deep neural network for supervised image classification, one can broadly distinguish between two types of latent features of images that will drive the classification of class Y . Following the notation of Gong et al. (2016), we can divide features broadly into the classes of (i) ‘core’ or ‘conditionally invariant’ features X whose distribution P (X|Y ) does not change substanti...

متن کامل

CyCADA: Cycle-Consistent Adversarial Domain Adaptation

Domain adaptation is critical for success in new, unseen environments. Adversarial adaptation models applied in feature spaces discover domain invariant representations, but are difficult to visualize and sometimes fail to capture pixel-level and low-level domain shifts. Recent work has shown that generative adversarial networks combined with cycle-consistency constraints are surprisingly effec...

متن کامل

Cycada: Cycle-consistent Adversarial Domain Adaptation

Domain adaptation is critical for success in new, unseen environments. Adversarial adaptation models applied in feature spaces discover domain invariant representations, but are difficult to visualize and sometimes fail to capture pixel-level and low-level domain shifts. Recent work has shown that generative adversarial networks combined with cycle-consistency constraints are surprisingly effec...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2017